Encoding Structured Output Values

نویسندگان

  • Edward Loper
  • Mitch Marcus
  • Fernando Pereira
  • Martha Palmer
چکیده

ENCODING STRUCTURED OUTPUT VALUES Edward Loper Martha Palmer Many of the Natural Language Processing tasks that we would like to model with machine learning techniques generate structured output values, such as trees, lists, or groupings. These structured output problems can be modeled by decomposing them into a set of simpler sub-problems, with well-defined and well-constrained interdependencies between sub-problems. However, the effectiveness of this approach depends to a large degree on exactly how the problem is decomposed into sub-problems; and on how those sub-problems are divided into equivalence classes. The notion of output encoding can be used to examine the effects of problem decomposition on learnability for specific tasks. These effects can be divided into two general classes: local effects and global effects. Local effects, which influence the difficulty of learning individual sub-problems, depend primarily on the coherence of the classes defined by individual output tags. Global effects, which determine the model’s ability to learn long-distance dependencies, depend on the information content of the output tags. Using a canonical encoding as a reference point, we can define additional encodings as reversible transformations from canonical encoded structures to a new set of encoded structures. This allows us to define a space of potential encodings (and by extension, a space of potential problem decompositions). Using search methods, we can then analyze and improve upon existing problem decompositions. For my dissertation, I plan to apply automatic and semi-automatic methods to the problem of finding optimal problem decompositions, in the context of three specific systems (one chunking system and two semantic role labeling systems). Additionally, I plan to evaluate a novel approach to voting between multiple models when each model uses a different problem decomposition, which I describe in Chapter 7.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Effective Bayesian Inference for Stochastic Programs

In this paper, we propose a stochastic version of a general purpose functional programming language as a method of modeling stochastic processes. The language contains random choices, conditional statements, structured values, defined functions, and recursion. By imagining an experiment in which the program is “run” and the random choices made by sampling, we can interpret a program in this lan...

متن کامل

1997-Effective Bayesian Inference for Stochastic Programs

In this paper, we propose a stochastic version of a general purpose functional programming language as a method of modeling stochastic processes. The language contains random choices, conditional statements, structured values, defined functions, and recursion. By imagining an experiment in which the program is “run” and the random choices made by sampling, we can interpret a program in this lan...

متن کامل

Refinement of learned skilled movement representation in motor cortex deep output layer

The mechanisms underlying the emergence of learned motor skill representation in primary motor cortex (M1) are not well understood. Specifically, how motor representation in the deep output layer 5b (L5b) is shaped by motor learning remains virtually unknown. In rats undergoing motor skill training, we detect a subpopulation of task-recruited L5b neurons that not only become more movement-encod...

متن کامل

Effects of Structured Input and Meaningful Output on EFL Learners' Acquisition of Nominal Clauses

The current second language (L2) instruction research has raised great motivation for the use of both processing instruction and meaningful output instruction tasks in L2 classrooms as the two focus-on-form (FonF) instructional tasks. The present study investigated the effect of structured input tasks (represented by referential and affective tasks) compared with meaningful output tasks (implem...

متن کامل

Constructive Reservoir Computation with Output Feedbacks for Structured Domains

We introduce a novel constructive algorithm which progressively builds the architecture of GraphESN, which generalizes Reservoir Computing to learning in graph domains. Exploiting output feedback signals in a forward fashion in such construction, allows us to introduce supervision in the reservoir encoding process. The potentiality of the proposed approach is experimentally assessed on real-wor...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2007